通过比较算法性能,公共攀爬的攀岩可以大大加速ML研究。但是,它需要与任务相关的指标。对于涉及寄生虫负荷的疾病,例如疟疾和被忽视的热带疾病(NTDS),例如血吸虫病,目前在ML论文中报道的指标(例如AUC,F1分数)不适合临床任务。结果,爬山系统并没有使解决这些严重疾病的解决方案取得进展。本文借鉴了疟疾和NTD的示例,在当前的ML实践中强调了两个差距,并提出了改进的方法:(i)我们描述了ML开发的方面,尤其是性能指标,需要将其牢固地基于临床用途案例。 ,我们提供获取此领域知识的方法。 (ii)我们详细描述了绩效指标,以指导涉及寄生虫负荷的疾病的ML模型的开发。我们强调了患者级别的观点,室内变异性,假阳性率,检测限制和不同类型的错误的重要性。我们还讨论了在这种情况下常用的ROC曲线和AUC的问题。
translated by 谷歌翻译
自动化数据驱动的建模,直接发现系统的管理方程的过程越来越多地用于科学界。 Pysindy是一个Python包,提供用于应用非线性动力学(SINDY)方法的稀疏识别到数据驱动模型发现的工具。在Pysindy的这一主要更新中,我们实现了几种高级功能,使得能够从嘈杂和有限的数据中发现更一般的微分方程。延长候选术语库,用于识别致动系统,部分微分方程(PDE)和隐式差分方程。还实施了包括Sindy和合奏技术的整体形式的强大配方,以提高现实世界数据的性能。最后,我们提供了一系列新的优化算法,包括多元稀疏的回归技术和算法来强制执行和促进不等式约束和稳定性。这些更新在一起,可以在文献中尚未报告的全新SINDY模型发现能力,例如约束PDE识别和使用不同稀疏的回归优化器合并。
translated by 谷歌翻译
我们考虑从高噪声限制的时间序列数据中控制方程的数据驱动发现。该算法开发描述了在非线性动力学(SINDY)框架的稀疏识别的背景下避免噪声的广泛影响的方法的广泛工具包。我们提供了两个主要贡献,都集中在系统x'= f(x)中获取的嘈杂数据。首先,我们提出用于高噪声设置的广泛工具包,这是一个批判性的回归方法的扩展,从完整的库中逐步剔除剔除功能,并产生一组稀疏方程,其回归到衍生x' 。这些创新可以从高噪声时间序列数据中提取稀疏控制方程和系数(例如,增加噪声300%)。例如,它发现洛伦茨系统中的正确稀疏文库,中值系数估计误差等于1% - 3%(50%噪声),6% - 8%(100%噪声);和23% - 25%(噪音300%)。工具包中的启用模块组合成单个方法,但各个模块可以在其他方程发现方法(Sindy或不)中进行战术,以改善高噪声数据的结果。其次,我们提出了一种技术,适用于基于X'= F(X)的任何模型发现方法,以评估由于噪声数据而在非唯一解决方案的上下文中发现模型的准确性。目前,这种非唯一性可以模糊发现模型的准确性,从而造成发现方法的有效性。我们描述了一种使用线性依赖性的技术,该技术将发现的模型转换为最接近真实模型的等效形式,从而能够更准确地评估发现的模型的准确性。
translated by 谷歌翻译
疟疾是一种威胁生命的疾病,影响了数百万。基于显微镜的薄膜评估是(i)确定疟疾物种和(ii)定量高寄生虫感染的标准方法。通过机器学习(ML)对疟疾显微镜的完全自动化是一项具有挑战性的任务,因为预先准备的滑动在质量和表现方面差异很大,并且伪像通常超过相对较少的寄生虫。在这项工作中,我们描述了一个用于薄膜疟疾分析的完整,完全自动化的框架,该框架应用了ML方法,包括卷积神经网(CNN),该方法在大型且多样化的田间预先准备的薄膜数据集中进行了训练。定量和物种鉴定结果几乎足够准确地满足了耐药性监测和临床用例的混凝土需求。我们将方法和性能指标集中在现场用例要求上。我们讨论了将ML方法应用于疟疾显微镜的关键问题和重要指标。
translated by 谷歌翻译
语言模型既展示了定量的改进,又展示了新的定性功能,随着规模的增加。尽管它们具有潜在的变革性影响,但这些新能力的特征却很差。为了为未来的研究提供信息,为破坏性的新模型能力做准备,并改善社会有害的效果,至关重要的是,我们必须了解目前和近乎未来的能力和语言模型的局限性。为了应对这一挑战,我们介绍了超越模仿游戏基准(Big Bench)。 Big Bench目前由204个任务组成,由132家机构的442位作者贡献。任务主题是多样的,从语言学,儿童发展,数学,常识性推理,生物学,物理学,社会偏见,软件开发等等。 Big-Bench专注于被认为超出当前语言模型的功能的任务。我们评估了OpenAI的GPT型号,Google内部密集变压器体系结构和大型基础上的开关稀疏变压器的行为,跨越了数百万到数十亿个参数。此外,一个人类专家评估者团队执行了所有任务,以提供强大的基准。研究结果包括:模型性能和校准都随规模改善,但绝对的术语(以及与评估者的性能相比);在模型类中的性能非常相似,尽管带有稀疏性。逐渐和预测的任务通常涉及大量知识或记忆成分,而在临界规模上表现出“突破性”行为的任务通常涉及多个步骤或组成部分或脆性指标;社交偏见通常会随着含糊不清的环境而随着规模而增加,但这可以通过提示来改善。
translated by 谷歌翻译
目的:基于知识的计划(KBP)通常涉及培训端到端深度学习模型以预测剂量分布。但是,由于经常使用的医疗数据集规模有限,端到端方法可能与实际限制有关。为了解决这些局限性,我们提出了一种基于内容的图像检索(CBIR)方法,用于根据解剖学相似性检索先前计划的患者的剂量分布。方法:我们提出的CBIR方法训练一种代表模型,该模型可产生患者解剖信息的潜在空间嵌入。然后将新患者的潜在空间嵌入与数据库中以前患者的潜在空间嵌入,以检索剂量分布的图像。该项目的所有源代码均可在GitHub上获得。主要结果:在由我们机构的公开计划和临床计划组成的数据集上评估了各种CBIR方法的检索性能。这项研究比较了各种编码方法,从简单的自动编码器到Simsiam等最新的暹罗网络,并且在Multipask Siamese网络中观察到了最佳性能。意义:应用CBIR告知后续的治疗计划可能会解决与端到端KBP相关的许多限制。我们目前的结果表明,可以通过对先前开发的暹罗网络进行轻微更改来获得出色的图像检索性能。我们希望通过Metaplanner框架等方法将CBIR集成到未来工作中的自动化计划工作流程中。
translated by 谷歌翻译
Federated learning (FL) is a machine learning setting where many clients (e.g. mobile devices or whole organizations) collaboratively train a model under the orchestration of a central server (e.g. service provider), while keeping the training data decentralized. FL embodies the principles of focused data collection and minimization, and can mitigate many of the systemic privacy risks and costs resulting from traditional, centralized machine learning and data science approaches. Motivated by the explosive growth in FL research, this paper discusses recent advances and presents an extensive collection of open problems and challenges.
translated by 谷歌翻译
Coronary Computed Tomography Angiography (CCTA) provides information on the presence, extent, and severity of obstructive coronary artery disease. Large-scale clinical studies analyzing CCTA-derived metrics typically require ground-truth validation in the form of high-fidelity 3D intravascular imaging. However, manual rigid alignment of intravascular images to corresponding CCTA images is both time consuming and user-dependent. Moreover, intravascular modalities suffer from several non-rigid motion-induced distortions arising from distortions in the imaging catheter path. To address these issues, we here present a semi-automatic segmentation-based framework for both rigid and non-rigid matching of intravascular images to CCTA images. We formulate the problem in terms of finding the optimal \emph{virtual catheter path} that samples the CCTA data to recapitulate the coronary artery morphology found in the intravascular image. We validate our co-registration framework on a cohort of $n=40$ patients using bifurcation landmarks as ground truth for longitudinal and rotational registration. Our results indicate that our non-rigid registration significantly outperforms other co-registration approaches for luminal bifurcation alignment in both longitudinal (mean mismatch: 3.3 frames) and rotational directions (mean mismatch: 28.6 degrees). By providing a differentiable framework for automatic multi-modal intravascular data fusion, our developed co-registration modules significantly reduces the manual effort required to conduct large-scale multi-modal clinical studies while also providing a solid foundation for the development of machine learning-based co-registration approaches.
translated by 谷歌翻译
Three main points: 1. Data Science (DS) will be increasingly important to heliophysics; 2. Methods of heliophysics science discovery will continually evolve, requiring the use of learning technologies [e.g., machine learning (ML)] that are applied rigorously and that are capable of supporting discovery; and 3. To grow with the pace of data, technology, and workforce changes, heliophysics requires a new approach to the representation of knowledge.
translated by 谷歌翻译
Image classification with small datasets has been an active research area in the recent past. However, as research in this scope is still in its infancy, two key ingredients are missing for ensuring reliable and truthful progress: a systematic and extensive overview of the state of the art, and a common benchmark to allow for objective comparisons between published methods. This article addresses both issues. First, we systematically organize and connect past studies to consolidate a community that is currently fragmented and scattered. Second, we propose a common benchmark that allows for an objective comparison of approaches. It consists of five datasets spanning various domains (e.g., natural images, medical imagery, satellite data) and data types (RGB, grayscale, multispectral). We use this benchmark to re-evaluate the standard cross-entropy baseline and ten existing methods published between 2017 and 2021 at renowned venues. Surprisingly, we find that thorough hyper-parameter tuning on held-out validation data results in a highly competitive baseline and highlights a stunted growth of performance over the years. Indeed, only a single specialized method dating back to 2019 clearly wins our benchmark and outperforms the baseline classifier.
translated by 谷歌翻译